Fill out this form to register your intent to complete this project.
Fill out this form to submit your solution to this project and qualify for the rewards.
![]() | Develop a fault-tolerant controller for a quadcopter using model-based reinforcement learning. |
Unmanned aerial vehicles (UAVs), such as quadcopters, are nowadays popular vehicles adopted for various applications. The actual operation of quadcopters can be affected by disturbances such as wind, rain, dust, etc., which may cause component faults. Fault is defined as a change in a system's property or parameters that cause the system to behave differently from its design. A fault-tolerant controller (FTC) is a control strategy that aims to improve the performance of a system that is operating in degraded performance due to fault [1].
Deployment of multi-rotor drones for applications like Urban Air Mobility (UAM), product delivery, etc., requires a proper behavior of the vechile at all times to ensure safety of the environemnt and people nearby. That is why a FTC is a necessary block for such systems and both academic researchers and industry professionals are looking to improve methods to observe faults and provide a control strategy able to overcome them and ensuring safe behavior.
FTCs are characterized as model-based or data-driven based on the method used to develop the controllers. Model-based techniques necessitate knowledge of the system's model and parameters in order to design a fault-tolerant controller. Data-driven approaches, on the other hand, learn the FTC directly from system data. The fundamental problem of model-based FTC approaches is that their effectiveness is dependent on the correctness of the system model, which is difficult to establish when system parameters vary due to faults. Furthermore, complex systems necessitate complicated controllers, which has an impact on the controllers' robustness. Data-driven techniques, on the other hand, utilize data to design FTC without knowing the full dynamics of the system. As a result, data-driven methods, particularly reinforcement learning (RL)-based techniques, have recently gained the attention of a number of researchers.
Train an RL agent to develop a fault-tolerant controller for a quadcopter using model-based reinforcement learning. The framework uses the system dynamics and a Kalman filter-based estimator to estimate the fault-related parameters online, which will be used to identify the occurrence of a fault in the system. Once you identify the event of a fault, you will use the fault-related parameters to train an RL agent that tunes the position and attitude controller gains of the quadcopter to compensate for the happening fault.
Suggested steps:
- Review Tune PI Controller using Reinforcement Learning example to learn how to use the Reinforcement Learning Toolbox to tune a PI controller for a system.
- Review Quadcopter Drone Model in Simscape example, which contains a detailed model of the quadcopter including the airframe, battery, and propulsion systems, and learn how a PID control can be applied for a quadcopter's position and attitude control.
- Design a reward function that will be used for training the RL agent (consider a reward function that takes into account the error between the reference and actual trajectory). To represent fault behavior of the system, you may use the equivalent resistance of the motors as in [2].
- Use the simulation environment to simulate faulty behaviors and train an RL agent to tune the quadcopter position and attitude PID controller gains.
- Apply the trained model for tuning the PID controllers in the presence of fault/s.
Advanced Project work:
- Implement a state estimator for monitoring the fault related parameters that will be used for training the RL agent (you may refer to Fault Detection Using an Extended Kalman Filter, [2], and [3]).
- Consider complete sub-component failure instead of fault (degradation).
Examples:
- PID Autotuning for UAV Quadcopter
- UAV Inflight Failure Recovery
- Simulink Drone Reference Application
- UAV Package Delivery
Suggested readings:
- [1] Blanke, M., Kinnaert, M., Lunze, J., Staroswiecki, M., and Schröder, J., Diagnosis and fault-tolerant control, Vol. 2, Springer, 2006.
- [2] Bhan, L., Quinones-Grueiro, M., and Biswas, G., “Fault Tolerant Control combining Reinforcement Learning and Model-based Control,” 2021 5th International Conference on Control and Fault-Tolerant Systems (SysTol), pp. 31–36, 2021.
- [3] Matthew Daigle, Bhaskar Saha, and Kai Goebel. A comparison of filter-based approaches for model based prognostics. In 2012 IEEE Aerospace Conference, pages 1–10, 2012.
Improve safety of multi-rotor drones
Drones, Artificial Intelligence, Robotics, Control, Reinforcement Learning, UAV
Master's, Doctoral
Dedicated discussion forum to ask/answer questions, comment, or share your ideas for solutions for this project.
235